On Boolean Combinations of Definitive Classifiers
نویسنده
چکیده
We consider the sample complexity of concept learning when we classify by using a fixed Boolean function of the outputs of a number of different classifiers. Here, we take into account the ‘margins’ of each of the constituent classifiers. A special case is that in which the constituent classifiers are linear threshold functions (or perceptrons) and the fixed Boolean function is the majority function. This corresponds to a ‘committee of perceptrons’, an artificial neural network (or circuit) consisting of a single layer of perceptrons (or linear threshold units) in which the output of the network is defined to be the majority output of the perceptrons. Recent work of Auer et al. studied the computational properties of such networks (where they were called ‘parallel perceptrons’), proposed an incremental learning algorithm for them, and demonstrated empirically that the learning rule is effective. As a corollary of the sample complexity result presented here, generalization error bounds are derived for this special case that provide further motivation for the use of this learning rule. ∗Department of Mathematics and Centre for Discrete and Applicable Mathematics, London School of Economics, London WC2A 2AE, UK. [email protected]
منابع مشابه
Iterative Boolean combination of classifiers in the ROC space: An application to anomaly detection with HMMs
Hidden Markov models (HMMs) have been shown to provide a high level performance for detecting anomalies in sequences of system calls to the operating system kernel. Using Boolean conjunction and disjunction functions to combine the responses of multiple HMMs in the ROC space may significantly improve performance over a ‘‘single best’’ HMM. However, these techniques assume that the classifiers a...
متن کاملMin-max classifiers: Learnability, design and application
This paper introduces the class of min max classifiers. These are binary-valued functions that can be used as pattern classifiers of both real-valued and binary-valued feature vectors. They are also lattice-theoretic generalization of Boolean functions and are also related to feed-forward neural networks and morphological signal operators. We studied supervised learning of these classifiers und...
متن کاملIncremental Boolean Combination of Classifiers
The incremental Boolean combination (incrBC ) technique is a new learn-and-combine approach that is proposed to adapt ensemblebased pattern classification systems over time, in response to new data acquired during operations. When a new block of training data becomes available, this technique generates a diversified pool of base classifiers from the data by varying training hyperparameters and ...
متن کاملNeural Algebra of Classifiers
The world is fundamentally compositional, so it is natural to think of visual recognition as the recognition of basic visually primitives that are composed according to well-defined rules. This strategy allows us to recognize unseen complex concepts from simple visual primitives. However, the current trend in visual recognition follows a data greedy approach where huge amounts of data are requi...
متن کاملFilter theory in MTL-algebras based on Uni-soft property
The notion of (Boolean) uni-soft filters in MTL-algebras is introduced, and several properties of them are investigated. Characterizations of (Boolean) uni-soft filters are discussed, and some (necessary and sufficient) conditions for a uni-soft filter to be Boolean are provided. The condensational property for a Boolean uni-soft filter is established.
متن کامل